It has been long speculated that expression of emotions from different modalities have the same underlying\r\nââ?¬Ë?codeââ?¬â?¢, whether it be a dance step, musical phrase, or tone of voice. This is the first attempt to implement this\r\ntheory across three modalities, inspired by the polyvalence and repeatability of robotics. We propose a unifying\r\nframework to generate emotions across voice, gesture, and music, by representing emotional states as a\r\n4-parameter tuple of speed, intensity, regularity, and extent (SIRE). Our results show that a simple 4-tuple can\r\ncapture four emotions recognizable at greater than chance across gesture and voice, and at least two emotions\r\nacross all three modalities. An application for multi-modal, expressive music robots is discussed.
Loading....